Generative image captioning in Urdu using deep learning
نویسندگان
چکیده
Abstract Urdu is morphologically rich language and lacks the resources available in English. While several studies on image captioning task English have been published, this among pioneer generative captioning. The study makes key contributions: (i) it presents a new dataset for captioning, (ii) different attention-based architectures language. These attention mechanisms are to language, as those never used (iii) Finally, performs quantitative qualitative analysis of results by studying impact model Urdu’s caption generation task. extensive experiments show encouraging such BLEU-1 score 72.5, BLEU-2 56.9, BLEU-3 42.8, BLEU-4 31.6. we present data code future research via GitHub ( https://github.com/saeedhas/Urdu_cap_gen ).
منابع مشابه
Deep Learning for Automatic Image Captioning in Poor Training Conditions
English. Recent advancements in Deep Learning show that the combination of Convolutional Neural Networks and Recurrent Neural Networks enables the definition of very effective methods for the automatic captioning of images. Unfortunately, this straightforward result requires the existence of large-scale corpora and they are not available for many languages. This paper describes a simple methodo...
متن کاملGenerative Choreography using Deep Learning
Recent advances in deep learning have enabled the extraction of high-level features from raw sensor data which has opened up new possibilities in many different fields, including computer generated choreography. In this paper we present a system chorrnn for generating novel choreographic material in the nuanced choreographic language and style of an individual choreographer. It also shows promi...
متن کاملLearning to Evaluate Image Captioning
Evaluation metrics for image captioning face two challenges. Firstly, commonly used metrics such as CIDEr, METEOR, ROUGE and BLEU often do not correlate well with human judgments. Secondly, each metric has well known blind spots to pathological caption constructions, and rulebased metrics lack provisions to repair such blind spots once identified. For example, the newly proposed SPICE correlate...
متن کاملContrastive Learning for Image Captioning
Image captioning, a popular topic in computer vision, has achieved substantial progress in recent years. However, the distinctiveness of natural descriptions is often overlooked in previous work. It is closely related to the quality of captions, as distinctive captions are more likely to describe images with their unique aspects. In this work, we propose a new learning method, Contrastive Learn...
متن کاملDeep Learning for Video Classification and Captioning
Accelerated by the tremendous increase in Internet bandwidth and storage space, video data has been generated, published and spread explosively, becoming an indispensable part of today's big data. In this paper, we focus on reviewing two lines of research aiming to stimulate the comprehension of videos with deep learning: video classification and video captioning. While video classification con...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of Ambient Intelligence and Humanized Computing
سال: 2023
ISSN: ['1868-5137', '1868-5145']
DOI: https://doi.org/10.1007/s12652-023-04584-y